Skip to content

Autoloops/upskill

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

27 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

upskill

Your AI agent works from memory. upskill makes it start from proven playbooks.
The skill layer that routes agents to the right workflow before they start real work.

npm version MIT license Browse upskill 10,000+ skills

upskill turns a web design task into a better first draft by injecting the right playbook

What is upskill?

upskill helps AI assistants use the right skill before they start working. It is a free, MIT-licensed routing layer for skills: the agent describes the task, upskill finds the best playbook, and the agent follows it instead of guessing from memory.

A skill is a proven playbook: instructions, examples, constraints, tools, and patterns for a specific kind of work. Instead of asking an agent to invent a pitch-deck structure, design system, inbox triage process, auth flow, research workflow, or browser automation script from memory, upskill finds the best existing playbook and puts it in context first.

Use it for serious work across code, docs, slides, email, research, spreadsheets, browser tasks, design, data, auth, cloud, CRM, support, and automation.

The expertise already exists: frontend design skills from Anthropic, implementation workflows from OpenAI, Stripe, Vercel, Microsoft, and others; curated practitioner skills from Garry Tan's gstack and obra/superpowers; and independent workflows from the community. The missing piece is routing the agent to the right one at the right time.

Quickstart

Paste this into Claude Code, Cursor, Codex, Cline, Windsurf, or any shell-capable AI assistant:

Install upskill for this assistant: run npm install -g @autoloops/upskill && upskill install; ask me four setup choices before changing config: telemetry on/off, local context/env-var names on/off, submissions on/off, and search scope verified/reviewed/community; apply my answers with upskill config set; run npx -y skills add Autoloops/upskill/skill; ask before adding the persistent rule; if I say yes, append the upskill rule to CLAUDE.md, AGENTS.md, .cursorrules, .clinerules, .windsurfrules, or ~/.claude/CLAUDE.md without overwriting anything.

For humans who want to run it directly:

npm install -g @autoloops/upskill
upskill install
upskill find "triage my inbox and surface what needs a reply today"

Why upskill?

AI agents are generalists. When they start from memory, they improvise.

The frustrating part is not that the model is bad. It is that the right answer often already exists somewhere: a frontend design playbook, an auth implementation guide, a CSV parsing pattern, a research workflow, a spreadsheet cleanup recipe. Your agent just does not know to reach for it.

Work Without upskill With upskill
Landing pages Generates a generic off-brand Tailwind hero Follows a frontend design playbook for layout, typography, motion, and accessibility
Clerk/auth Skips session checks, callbacks, or JWT verification details Follows a provider-specific auth flow with the expected edge cases
CSV/data parsing Reinvents half of papaparse, badly Uses the right library and edge-case checklist
Pitch decks Produces a generic template Follows a narrative arc and slide-quality rubric
Email Lists unread messages Builds a prioritized action queue
Research Summarizes loosely Produces a cited synthesis with gaps and sources
UI Generates generic layouts Uses a design and component playbook
Browser tasks Clicks through fragile selectors Uses a tested automation workflow

The result: fewer retries, less token waste, and better output on the first pass.

The expertise exists. The routing does not. upskill adds that missing layer.

Demo

Task: Make me a polished 12-slide seed deck as an editable PPTX

Without upskill, an assistant usually starts from a generic deck outline:

1. Title
2. Problem
3. Solution
4. Market
5. Product
6. Business model
7. Team
8. Ask

The slides look familiar because the agent is guessing from memory: weak narrative, inconsistent visuals, no speaker notes, and no real investor-quality review pass.

With upskill, the assistant can start from a deck-writing and PPTX playbook:

upskill find "create a polished seed pitch deck as an editable pptx"
upskill inspect <pitch-deck-or-pptx-skill>

Then it follows the playbook:

Deck part What changes with upskill
Narrative Hook, problem, insight, solution, proof, market, GTM, ask
Slide quality One idea per slide, stronger hierarchy, less filler text
Visual system Consistent type, spacing, color, charts, and layout rules
PPTX output Editable slides instead of a throwaway text outline
Review pass Checks story gaps, weak claims, crowded slides, and missing evidence

The difference is simple: the assistant got the right skill before it started.

How it works

  1. Search — the assistant runs upskill find "<task>".
  2. Inspect — it reads the best matching skill before execution.
  3. Apply — it follows the proven playbook instead of going freehand.
  4. Improve — if you opted in, it reports whether the skill worked.
upskill find "turn this customer feedback spreadsheet into the top 5 product themes"
upskill inspect <skill_id>

Once the assistant skill is installed, your agent can do this automatically before non-trivial tasks.

Examples

"Make me a 12-slide pitch deck"

upskill can surface a deck-writing skill with a narrative structure, slide order, quality bar, and review checklist, so the assistant does not produce another generic template.

"Triage my inbox"

upskill can surface an email triage playbook: classify action/FYI/noise, rank by sender and urgency signals, and return only what needs attention today.

"Research competitors"

upskill can surface a research workflow that separates claims from evidence, tracks sources, and produces a structured comparison instead of a loose summary.

"Add auth to this app"

upskill can surface provider-specific setup guidance, expected env vars, scopes, callbacks, and implementation pitfalls before the assistant writes code.

"Design a landing page"

upskill can surface a frontend design skill with layout, typography, visual hierarchy, motion, accessibility, and design-review constraints, so the assistant does not produce the same generic hero again.

"Automate a Google Workspace task"

upskill can surface a workflow that knows the difference between Gmail, Calendar, Drive, Docs, Sheets, and the auth path each one needs.

Not just code

upskill is for any serious agent work where there is a better playbook than winging it:

  • slides and pitch decks
  • email triage
  • Google Workspace workflows
  • Notion and knowledge-base queries
  • calendar automation
  • scientific writing
  • accessibility audits
  • malware and security analysis
  • sales and support playbooks
  • browser workflows
  • cloud, auth, and developer tools

Trust and control

upskill is designed so the user stays in control.

Default Behavior
Verified search Searches trusted sources first
Telemetry off No outcome reporting unless enabled
Context sharing off No local environment context unless enabled
Env values protected Context sharing sends env-var names only, never values
Submissions off No publishing unless enabled and confirmed
Rule approval Persistent assistant rules are appended only after user approval

Inspect settings anytime:

upskill config show

Change settings anytime:

upskill config set telemetry true
upskill config set context true
upskill config set submissions true
upskill config set search-scope verified

Technical architecture

upskill has three pieces:

Piece What it does
CLI Lets an assistant search, inspect, report outcomes, submit skills, and manage local privacy settings.
Assistant skill Teaches Claude Code, Cursor, Codex, Cline, Windsurf, and similar agents to call upskill before non-trivial work.
Registry Stores indexed skills, trust metadata, search vectors, source info, feedback stats, and compatibility signals.

The core loop is intentionally simple:

  1. A skill is indexed from a public source or submitted source.
  2. The registry stores the skill text plus derived metadata: task tags, dependencies, auth requirements, env-var names, commands, permissions, warnings, trust level, source URL, and source commit.
  3. An agent calls upskill find "<task>".
  4. The registry returns ranked skills with match explanations and missing requirements.
  5. The agent calls upskill inspect <skill_id> and reads the full SKILL.md.
  6. The agent follows the skill instead of improvising.
  7. If telemetry is enabled, the agent reports whether the skill worked.

This is the important distinction: upskill is not trying to be another chat UI. It is a skill-selection layer that gives agents better context before execution.

Think of it as mixture of experts at the agent layer: the model stays general, but the task gets routed to a specialized playbook before the agent acts.

Security model

upskill is designed around inspection, pinning, and small payloads.

Control Behavior
Trust tiers Search can be limited to verified, expanded to reviewed, or opened to community.
Pinned sources Skills resolve to source metadata, including GitHub URL/path/ref data where available, so agents can inspect what they are using.
No auto-install find and inspect return instructions and source info. They do not silently install code or mutate your project.
Human-readable skills The execution contract is SKILL.md, not an opaque binary package. Agents can inspect the instructions before following them.
Outbound safety checks CLI payloads are scanned for known secret patterns before feedback or submissions leave the machine.
Submission guardrails Local folder submissions are capped and scanned for secret-looking files and values before upload.
Local opt-ins Telemetry, environment context, and submissions are disabled until explicitly enabled.

For untrusted or community skills, the registry can apply heavier review before promotion: dependency extraction, auth detection, env-var detection, dangerous command checks, network/secret access warnings, and LLM-assisted security review. The goal is not to pretend every public skill is safe. The goal is to surface trust, requirements, and warnings before an agent uses it.

The review layer is built to catch practical problems agents are likely to miss: prompt injection, credential exfiltration, typosquatting, lookalike domains, hidden malicious instructions, unsafe shell patterns, destructive file operations, and suspicious network access. Examples include hidden HTML/script payloads, instructions that tell the agent to skip verification, packages that look like trusted dependencies, and commands that read secrets before making network calls. Some findings will be warnings rather than hard blocks because legitimate workflows can still contain risky-looking operations, such as deleting node_modules or calling a delete endpoint.

Vetted, fresh, stack-aware

Good recommendations need more than keyword search.

Signal Why it matters
Vetted sources Vendor-official and curated sources should rank above random public submissions.
Fresh indexing Skills can be updated faster than model training data, so agents get newer workflows.
Environment fit If context sharing is enabled, skills that match installed CLIs and available env-var names can rank higher.
Auth fit A Slack skill that needs Composio, a GitHub skill that needs gh auth, and an Exa skill that needs EXA_API_KEY should not be treated as interchangeable.
Feedback loop Skills that work should rise. Skills that fail for real agents should sink.

Example: if an agent has gh installed and GitHub auth available, a GitHub PR review skill that uses the GitHub CLI is a better recommendation than a generic code-review prompt. If the agent has COMPOSIO_API_KEY, broker-based Gmail or Slack skills become more useful than skills that require a manual OAuth setup.

Concrete examples:

  • AWS_ACCESS_KEY_ID exists locally → AWS deployment and cloud-operation skills can rank higher.
  • STRIPE_SECRET_KEY exists locally → Stripe-specific checkout and webhook flows can rank higher.
  • COMPOSIO_API_KEY exists locally → Composio-backed Gmail, Slack, Notion, and calendar skills can rank higher.
  • gh is installed and authenticated → GitHub PR, issue, and repo-maintenance skills can rank higher.

Only names are used for this matching. Values never leave your machine.

Ranking signals

Every result includes a match explanation so agents do not have to trust a black box.

Signal Meaning
name_match Query terms match the skill name. This is one of the strongest signals.
text Postgres full-text keyword overlap over the skill name, description, tags, and indexed text.
vec 1024-dim semantic/vector similarity when embeddings are available. Useful for paraphrases.
trust verified > reviewed > community.
Environment fit Required commands, package managers, runtimes, MCP servers, and env-var names match the agent's environment.
Missing requirements Missing auth, env vars, commands, or package managers reduce usefulness.
Warnings Risky commands, unclear auth, weak descriptions, or ambiguous tasks can lower confidence.
Feedback Successes, failures, and workaround codes from prior runs improve future ranking.
Popularity Installs, GitHub stars, forks, and freshness are tie-breakers, not the core ranking signal.

A strong hit usually has either a literal name match or a combination of high text relevance, semantic relevance, trust, and environment fit. The registry should explain why a skill ranked, not just return a mysterious score.

Hybrid search matters because both extremes fail:

  • pure vector search can miss exact details like CLI flags, API names, env vars, and provider-specific terms
  • pure full-text search can miss intent when the user phrases the task differently from the skill
  • hybrid search catches both the specific words and the underlying task

Privacy and data flow

Default behavior is intentionally small.

All registry requests include an anonymous local install ID, CLI version, and user agent so the server can handle client registration and compatibility. That ID is not your name, email, repo, or workspace path.

Action What is sent
upskill find The search query and configured search scope. If context sharing is enabled, it may also send installed command names and env-var names.
upskill inspect The skill ID being fetched.
upskill report Only if telemetry is enabled: skill version ID, outcome, task kind, failure codes, and workaround codes.
upskill submit Only if submissions are enabled: the skill folder or source reference after local safety checks.

What is not sent by default:

  • env-var values
  • raw logs
  • private source code
  • shell history
  • file paths from your project
  • identifying user data for outcome telemetry

Self-hosted or private registries can be used by setting:

UPSKILL_URL=https://your-registry.example.com
upskill config set server https://your-registry.example.com

Why this solves the problem

Models are broad generalists. Serious work needs task-specific process.

The usual agent failure mode is not that the model cannot write text or code. It is that it starts from vague memory, skips the boring edge cases, misses tool-specific setup, and repeats patterns that were common in training data but wrong for the current task.

upskill changes the starting point:

Without upskill With upskill
Agent guesses the workflow Agent starts from a proven workflow
Agent invents requirements Agent sees dependencies, auth, commands, and warnings
Agent repeats generic patterns Agent uses task-specific examples and constraints
Agent learns nothing from failure Outcome feedback improves future ranking
Human has to remember the right docs Agent pulls the right skill before execution

The long-term bet is simple: a skill registry should become for AI agents what package registries became for programmers. Agents should not reinvent common work every time.

What's next

  • Approval workflow: clearer promotion paths from submitted skills to reviewed and verified skills.
  • Richer metadata extraction: stronger detection for task fit, auth, dependencies, side effects, and required tools.
  • Embeddings everywhere: better semantic search over skill descriptions, summaries, examples, and task tags.
  • Registry-hosted distributions: support sources beyond GitHub while keeping hashes and provenance.
  • Verified authorship: stronger proof that official vendor skills actually come from the vendor.
  • Company registries: private skill registries for teams that need internal workflows behind a firewall.
  • Better feedback stats: compatibility by agent, OS, tools, auth path, and failure mode.

CLI

upskill find "build a clean 12 slide seed pitch deck"
upskill find "parse uploaded CSV files with headers and quoted fields"
upskill find "research competitors and produce a cited comparison"
upskill inspect <skill_id>
upskill config show

Contribute skills

If you have a workflow that reliably makes an assistant better, turn it into a skill.

Good skills are not clever prompts. They are reusable work patterns:

  • how to triage an inbox
  • how to build a pitch deck
  • how to review a pull request
  • how to query a knowledge base
  • how to parse messy files
  • how to automate a browser workflow
  • how to research with citations
  • how to follow a product or design standard
  • how to run a Google Workspace workflow
  • how to automate calendar operations
  • how to write or review scientific work
  • how to run accessibility audits
  • how to follow a sales or support playbook
  • how to do malware or security analysis safely

The goal is simple: every agent should start important work with the best available playbook.

License

MIT.

About

CLI + skill for the Autoloops upskill registry. Search, inspect, report on, and publish agent skills from your shell.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors